2025-04-03 22:46:14,198 [ 89193 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-04-03 22:46:14,199 [ 89193 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:97, check_args_and_update_paths) 2025-04-03 22:46:14,199 [ 89193 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:108, check_args_and_update_paths) 2025-04-03 22:46:14,199 [ 89193 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:110, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_sbt52f --privileged --dns-search='.' --memory=30709035008 --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=6712d5cc610d -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e CLICKHOUSE_USE_OLD_ANALYZER=1 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_accept_invalid_certificate/test.py::test_accept test_accept_invalid_certificate/test.py::test_connection_accept test_accept_invalid_certificate/test.py::test_default test_accept_invalid_certificate/test.py::test_strict_connection_reject test_accept_invalid_certificate/test.py::test_strict_reject test_accept_invalid_certificate/test.py::test_strict_reject_with_config test_always_fetch_merged/test.py::test_replica_always_download test_async_insert_memory/test.py::test_memory_usage test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_compression_nested_columns/test.py::test_nested_compression_codec test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_config_decryption/test_wrong_settings.py::test_invalid_chars test_config_decryption/test_wrong_settings.py::test_no_encryption_key test_config_decryption/test_wrong_settings.py::test_subnodes test_config_decryption/test_wrong_settings.py::test_wrong_method test_config_xml_main/test.py::test_xml_main_conf test_config_yaml_main/test.py::test_yaml_main_conf test_create_query_constraints/test.py::test_create_query_const_constraints test_create_query_constraints/test.py::test_create_query_minmax_constraints test_custom_settings/test.py::test_custom_settings test_custom_settings/test.py::test_illformed_setting test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica test_ddl_worker_replicas/test.py::test_ddl_worker_replicas test_default_role/test.py::test_alter_user test_default_role/test.py::test_set_default_roles test_default_role/test.py::test_wrong_set_default_role 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1]' test_dictionaries_dependency/test.py::test_no_lazy_load 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache]' test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation test_disk_access_storage/test.py::test_alter -vvv" altinityinfra/integration-tests-runner:cd6390247eca '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: random-0.2, timeout-2.2.0, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [100 items] scheduling tests via LoadFileScheduling test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] test_attach_partition_using_copy/test.py::test_all_replicated test_default_role/test.py::test_alter_user test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] test_config_decryption/test_wrong_settings.py::test_invalid_chars test_cluster_all_replicas/test.py::test_cluster test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] test_backup_restore/test.py::test_attach_partition test_accept_invalid_certificate/test.py::test_accept [gw7] [ 1%] PASSED test_config_decryption/test_wrong_settings.py::test_invalid_chars test_config_decryption/test_wrong_settings.py::test_no_encryption_key [gw1] [ 2%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] [gw1] [ 3%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] [gw4] [ 4%] PASSED test_accept_invalid_certificate/test.py::test_accept test_accept_invalid_certificate/test.py::test_connection_accept [gw4] [ 5%] PASSED test_accept_invalid_certificate/test.py::test_connection_accept test_accept_invalid_certificate/test.py::test_default [gw4] [ 6%] PASSED test_accept_invalid_certificate/test.py::test_default test_accept_invalid_certificate/test.py::test_strict_connection_reject [gw1] [ 7%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] [gw4] [ 8%] PASSED test_accept_invalid_certificate/test.py::test_strict_connection_reject test_accept_invalid_certificate/test.py::test_strict_reject [gw4] [ 9%] PASSED test_accept_invalid_certificate/test.py::test_strict_reject test_accept_invalid_certificate/test.py::test_strict_reject_with_config [gw4] [ 10%] PASSED test_accept_invalid_certificate/test.py::test_strict_reject_with_config [gw1] [ 11%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] [gw9] [ 12%] PASSED test_default_role/test.py::test_alter_user test_default_role/test.py::test_set_default_roles [gw1] [ 13%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] [gw1] [ 14%] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] [gw9] [ 15%] PASSED test_default_role/test.py::test_set_default_roles test_default_role/test.py::test_wrong_set_default_role test_concurrent_queries_restriction_by_query_kind/test.py::test_insert [gw9] [ 16%] PASSED test_default_role/test.py::test_wrong_set_default_role [gw1] [ 17%] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] [gw8] [ 18%] PASSED test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition [gw5] [ 19%] PASSED test_cluster_all_replicas/test.py::test_cluster test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] [gw0] [ 20%] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] [gw1] [ 21%] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] test_create_query_constraints/test.py::test_create_query_const_constraints test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] [gw7] [ 22%] PASSED test_config_decryption/test_wrong_settings.py::test_no_encryption_key test_config_decryption/test_wrong_settings.py::test_subnodes test_custom_settings/test.py::test_custom_settings [gw0] [ 23%] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] [gw8] [ 24%] PASSED test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] [gw5] [ 25%] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] [gw3] [ 26%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] [gw0] [ 27%] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica [gw8] [ 28%] PASSED test_backup_restore/test.py::test_restore test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] [gw7] [ 29%] PASSED test_config_decryption/test_wrong_settings.py::test_subnodes test_config_decryption/test_wrong_settings.py::test_wrong_method [gw5] [ 30%] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] test_cluster_all_replicas/test.py::test_global_in [gw4] [ 31%] PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select [gw1] [ 32%] PASSED test_custom_settings/test.py::test_custom_settings test_custom_settings/test.py::test_illformed_setting [gw5] [ 33%] PASSED test_cluster_all_replicas/test.py::test_global_in test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] [gw1] [ 34%] PASSED test_custom_settings/test.py::test_illformed_setting [gw0] [ 35%] PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica [gw9] [ 36%] PASSED test_create_query_constraints/test.py::test_create_query_const_constraints test_create_query_constraints/test.py::test_create_query_minmax_constraints test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_always_fetch_merged/test.py::test_replica_always_download test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication [gw9] [ 37%] PASSED test_create_query_constraints/test.py::test_create_query_minmax_constraints [gw0] [ 38%] PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table [gw5] [ 39%] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] [gw0] [ 40%] PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_async_insert_memory/test.py::test_memory_usage test_backup_restore_on_cluster/test.py::test_file_deduplication [gw2] [ 41%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] [gw0] [ 42%] PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication [gw7] [ 43%] PASSED test_config_decryption/test_wrong_settings.py::test_wrong_method test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field [gw3] [ 44%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] test_backup_restore_on_cluster/test.py::test_get_error_from_other_host [gw0] [ 45%] PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host [gw5] [ 46%] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] [gw0] [ 47%] PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size [gw4] [ 48%] PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_backup_restore_on_cluster/test.py::test_mutation test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] [gw0] [ 49%] PASSED test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection [gw3] [ 50%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] [gw8] [ 51%] PASSED test_always_fetch_merged/test.py::test_replica_always_download [gw0] [ 52%] PASSED test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database [gw7] [ 53%] PASSED test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] test_compression_nested_columns/test.py::test_nested_compression_codec [gw0] [ 54%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database test_config_xml_main/test.py::test_xml_main_conf test_backup_restore_on_cluster/test.py::test_replicated_database_async [gw2] [ 55%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] [gw3] [ 56%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] [gw5] [ 57%] PASSED test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica [gw0] [ 58%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] [gw0] [ 59%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] [gw7] [ 60%] PASSED test_config_xml_main/test.py::test_xml_main_conf [gw8] [ 61%] PASSED test_compression_nested_columns/test.py::test_nested_compression_codec [gw0] [ 62%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] [gw3] [ 63%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation [gw0] [ 64%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] [gw9] [ 65%] PASSED test_async_insert_memory/test.py::test_memory_usage test_backup_restore_on_cluster/test.py::test_replicated_table [gw0] [ 66%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table test_ddl_worker_replicas/test.py::test_ddl_worker_replicas [gw4] [ 67%] PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster [gw3] [ 68%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] [gw0] [ 69%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster [gw2] [ 70%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] [gw0] [ 71%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def [gw0] [ 72%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert [gw8] [ 73%] PASSED test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation [gw0] [ 74%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert [gw2] [ 75%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge [gw3] [ 76%] PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] [gw0] [ 77%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] [gw0] [ 78%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges [gw4] [ 79%] PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] [gw0] [ 80%] PASSED test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup [gw3] [ 81%] PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] test_dictionaries_dependency/test.py::test_no_lazy_load [gw0] [ 82%] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup [gw2] [ 83%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] test_backup_restore_on_cluster/test.py::test_system_functions [gw0] [ 84%] PASSED test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users [gw0] [ 85%] PASSED test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def [gw0] [ 86%] PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty [gw0] [ 87%] PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency [gw0] [ 88%] PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency [gw2] [ 89%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] [gw9] [ 90%] PASSED test_ddl_worker_replicas/test.py::test_ddl_worker_replicas [gw3] [ 91%] PASSED test_dictionaries_dependency/test.py::test_no_lazy_load [gw1] [ 92%] PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication [gw2] [ 93%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] test_config_yaml_main/test.py::test_yaml_main_conf [gw2] [ 94%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] [gw1] [ 95%] PASSED test_config_yaml_main/test.py::test_yaml_main_conf test_disk_access_storage/test.py::test_alter [gw1] [ 96%] PASSED test_disk_access_storage/test.py::test_alter [gw6] [ 97%] FAILED test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree [gw6] [ 98%] FAILED test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk [gw6] [ 99%] FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated [gw6] [100%] FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated =================================== FAILURES =================================== _____________________________ test_all_replicated ______________________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_all_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", True) test_attach_partition_using_copy/test.py:126: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ---------------------------- Captured stderr setup ----------------------------- Command:[docker ps | wc -l] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_attach_partition_using_copy/test.py Cluster start called. is_up=False Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestattachpartitionusingcopy-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: replica1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/database Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: replica2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/database Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:6712d5cc610d', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml pull] Stderr: zoo1 Skipped - Image is already being pulled by replica1 Stderr: zoo2 Skipped - Image is already being pulled by replica1 Stderr: zoo3 Skipped - Image is already being pulled by replica1 Stderr: replica2 Skipped - Image is already being pulled by replica1 Stderr: replica1 Pulling Stderr: replica1 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'] Command:[docker compose --project-name roottestattachpartitionusingcopy-gw6 --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr:time="2025-04-03T22:46:31Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestattachpartitionusingcopy-gw6_default Creating Stderr: Network roottestattachpartitionusingcopy-gw6_default Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Started Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Started Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Started Stderr:time="2025-04-03T22:46:32Z" level=debug msg="otel error" error="" Stderr:time="2025-04-03T22:46:32Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.3.2, port:2181, use_ssl:False Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.3.4, port:2181, use_ssl:False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.3.3, port:2181, use_ssl:False Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate] Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Running Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Running Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Running Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Started Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Started ClickHouse instance created get_instance_ip instance_name=replica1 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=replica1 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in replica1, ip: 172.16.3.6... http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None ClickHouse replica1 started get_instance_ip instance_name=replica2 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=replica2 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in replica2, ip: 172.16.3.5... http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7dfc5ee7066e31899e1c88eb764173812c150707dc23941009d5de56ab3a8d83/json HTTP/1.1" 200 None ClickHouse replica2 started ------------------------------ Captured log setup ------------------------------ 2025-04-03 22:46:20 [ 680 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:122, run_and_check) 2025-04-03 22:46:20 [ 680 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-03 22:46:20 [ 680 ] DEBUG : No running containers (conftest.py:96, cleanup_environment) 2025-04-03 22:46:20 [ 680 ] DEBUG : Pruning Docker networks (conftest.py:98, cleanup_environment) 2025-04-03 22:46:20 [ 680 ] DEBUG : Command:[docker network prune --force] (cluster.py:122, run_and_check) 2025-04-03 22:46:20 [ 680 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:122, run_and_check) 2025-04-03 22:46:20 [ 680 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:146, run_and_check) 2025-04-03 22:46:20 [ 680 ] INFO : Running tests in /ClickHouse/tests/integration/test_attach_partition_using_copy/test.py (cluster.py:2793, start) 2025-04-03 22:46:20 [ 680 ] DEBUG : Cluster start called. is_up=False (cluster.py:2800, start) 2025-04-03 22:46:20 [ 680 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-03 22:46:20 [ 680 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-03 22:46:20 [ 680 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-03 22:46:20 [ 680 ] DEBUG : Cleanup called (cluster.py:894, cleanup) 2025-04-03 22:46:20 [ 680 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-03 22:46:20 [ 680 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-03 22:46:20 [ 680 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-03 22:46:20 [ 680 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:122, run_and_check) 2025-04-03 22:46:20 [ 680 ] DEBUG : Unstopped containers: {} (cluster.py:908, cleanup) 2025-04-03 22:46:20 [ 680 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy-gw6 (cluster.py:922, cleanup) 2025-04-03 22:46:20 [ 680 ] DEBUG : Trying to prune unused networks... (cluster.py:928, cleanup) 2025-04-03 22:46:20 [ 680 ] DEBUG : Trying to prune unused images... (cluster.py:944, cleanup) 2025-04-03 22:46:20 [ 680 ] DEBUG : Command:[docker image prune -f] (cluster.py:122, run_and_check) 2025-04-03 22:46:20 [ 680 ] DEBUG : Stderr:Error response from daemon: a prune operation is already running (cluster.py:148, run_and_check) 2025-04-03 22:46:20 [ 680 ] DEBUG : Exitcode:1 (cluster.py:150, run_and_check) 2025-04-03 22:46:20 [ 680 ] DEBUG : Trying to prune unused volumes... (cluster.py:953, cleanup) 2025-04-03 22:46:20 [ 680 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:122, run_and_check) 2025-04-03 22:46:20 [ 680 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-03 22:46:20 [ 680 ] DEBUG : Volumes pruned: 1 (cluster.py:958, cleanup) 2025-04-03 22:46:20 [ 680 ] DEBUG : Setup directory for instance: replica1 (cluster.py:2813, start) 2025-04-03 22:46:20 [ 680 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4639, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Create directory for common tests configuration (cluster.py:4644, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Copy common configuration from helpers (cluster.py:4664, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Generate and write macros file (cluster.py:4716, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/configs/config.d (cluster.py:4752, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/database (cluster.py:4769, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs (cluster.py:4780, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Setup directory for instance: replica2 (cluster.py:2813, start) 2025-04-03 22:46:20 [ 680 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4639, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Create directory for common tests configuration (cluster.py:4644, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Copy common configuration from helpers (cluster.py:4664, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Generate and write macros file (cluster.py:4716, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/configs/config.d (cluster.py:4752, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/database (cluster.py:4769, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs (cluster.py:4780, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2025-04-03 22:46:20 [ 680 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:6712d5cc610d', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env (cluster.py:97, _create_env_file) 2025-04-03 22:46:20 [ 680 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-03 22:46:20 [ 680 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-03 22:46:20 [ 680 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-03 22:46:20 [ 680 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-03 22:46:20 [ 680 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2025-04-03 22:46:20 [ 680 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml pull] (cluster.py:122, run_and_check) 2025-04-03 22:46:31 [ 680 ] DEBUG : Stderr: zoo1 Skipped - Image is already being pulled by replica1 (cluster.py:148, run_and_check) 2025-04-03 22:46:31 [ 680 ] DEBUG : Stderr: zoo2 Skipped - Image is already being pulled by replica1 (cluster.py:148, run_and_check) 2025-04-03 22:46:31 [ 680 ] DEBUG : Stderr: zoo3 Skipped - Image is already being pulled by replica1 (cluster.py:148, run_and_check) 2025-04-03 22:46:31 [ 680 ] DEBUG : Stderr: replica2 Skipped - Image is already being pulled by replica1 (cluster.py:148, run_and_check) 2025-04-03 22:46:31 [ 680 ] DEBUG : Stderr: replica1 Pulling (cluster.py:148, run_and_check) 2025-04-03 22:46:31 [ 680 ] DEBUG : Stderr: replica1 Pulled (cluster.py:148, run_and_check) 2025-04-03 22:46:31 [ 680 ] DEBUG : Setup ZooKeeper (cluster.py:2854, start) 2025-04-03 22:46:31 [ 680 ] DEBUG : Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'] (cluster.py:2855, start) 2025-04-03 22:46:31 [ 680 ] DEBUG : Command:[docker compose --project-name roottestattachpartitionusingcopy-gw6 --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] (cluster.py:122, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr:time="2025-04-03T22:46:31Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Creating (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Created (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Creating (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Creating (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Creating (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Created (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Created (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Created (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Starting (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Starting (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Starting (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Started (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Started (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Started (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr:time="2025-04-03T22:46:32Z" level=debug msg="otel error" error="" (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Stderr:time="2025-04-03T22:46:32Z" level=debug msg="otel error" error="" (cluster.py:148, run_and_check) 2025-04-03 22:46:32 [ 680 ] DEBUG : Wait ZooKeeper to start (cluster.py:2466, wait_zookeeper_to_start) 2025-04-03 22:46:32 [ 680 ] DEBUG : get_instance_ip instance_name=zoo1 (cluster.py:2082, get_instance_ip) 2025-04-03 22:46:32 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:32 [ 680 ] DEBUG : get_kazoo_client: zoo1, ip:172.16.3.2, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-03 22:46:32 [ 680 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 22:46:32 [ 680 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 22:46:32 [ 680 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 22:46:32 [ 680 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 22:46:32 [ 680 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 22:46:32 [ 680 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 22:46:33 [ 680 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 22:46:33 [ 680 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 22:46:34 [ 680 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 22:46:34 [ 680 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 22:46:36 [ 680 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 22:46:36 [ 680 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-03 22:46:36 [ 680 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-03 22:46:36 [ 680 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-03 22:46:36 [ 680 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-03 22:46:36 [ 680 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-03 22:46:36 [ 680 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-03 22:46:36 [ 680 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-03 22:46:36 [ 680 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-03 22:46:36 [ 680 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-03 22:46:36 [ 680 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-03 22:46:36 [ 680 ] DEBUG : get_instance_ip instance_name=zoo2 (cluster.py:2082, get_instance_ip) 2025-04-03 22:46:36 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:36 [ 680 ] DEBUG : get_kazoo_client: zoo2, ip:172.16.3.4, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-03 22:46:36 [ 680 ] INFO : Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 22:46:36 [ 680 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-03 22:46:36 [ 680 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-03 22:46:36 [ 680 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-03 22:46:36 [ 680 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-03 22:46:36 [ 680 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-03 22:46:36 [ 680 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-03 22:46:36 [ 680 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-03 22:46:36 [ 680 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-03 22:46:36 [ 680 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-03 22:46:36 [ 680 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-03 22:46:36 [ 680 ] DEBUG : get_instance_ip instance_name=zoo3 (cluster.py:2082, get_instance_ip) 2025-04-03 22:46:36 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo3-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:36 [ 680 ] DEBUG : get_kazoo_client: zoo3, ip:172.16.3.3, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-03 22:46:36 [ 680 ] INFO : Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 22:46:36 [ 680 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-03 22:46:36 [ 680 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-03 22:46:36 [ 680 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-03 22:46:36 [ 680 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-03 22:46:36 [ 680 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-03 22:46:36 [ 680 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-03 22:46:36 [ 680 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-03 22:46:36 [ 680 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-03 22:46:36 [ 680 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-03 22:46:36 [ 680 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-03 22:46:36 [ 680 ] DEBUG : All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') (cluster.py:2482, wait_zookeeper_nodes_to_start) 2025-04-03 22:46:36 [ 680 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate') (cluster.py:3200, start) 2025-04-03 22:46:36 [ 680 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate] (cluster.py:122, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Running (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Running (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Running (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Creating (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Creating (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Created (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Created (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Starting (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Starting (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Started (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Started (cluster.py:148, run_and_check) 2025-04-03 22:46:37 [ 680 ] DEBUG : ClickHouse instance created (cluster.py:3208, start) 2025-04-03 22:46:37 [ 680 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2082, get_instance_ip) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2092, get_instance_global_ipv6) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : Waiting for ClickHouse start in replica1, ip: 172.16.3.6... (cluster.py:3216, start) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:37 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:38 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:38 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:38 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:38 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/03bec8185a6cda8aed878798e02a620f5d300cb05e341ada700f7b2c0cf549d8/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:38 [ 680 ] DEBUG : ClickHouse replica1 started (cluster.py:3220, start) 2025-04-03 22:46:38 [ 680 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2082, get_instance_ip) 2025-04-03 22:46:38 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:38 [ 680 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2092, get_instance_global_ipv6) 2025-04-03 22:46:38 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:38 [ 680 ] DEBUG : Waiting for ClickHouse start in replica2, ip: 172.16.3.5... (cluster.py:3216, start) 2025-04-03 22:46:38 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:38 [ 680 ] DEBUG : http://localhost:None "GET /v1.46/containers/7dfc5ee7066e31899e1c88eb764173812c150707dc23941009d5de56ab3a8d83/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 22:46:38 [ 680 ] DEBUG : ClickHouse replica2 started (cluster.py:3220, start) ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (139793411016256) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-03 22:46:38 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-03 22:46:38 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-03 22:46:38 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-03 22:46:39 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-03 22:46:39 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:47:33 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:48:27 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:49:23 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:50:20 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:51:17 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:52:12 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:53:07 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:54:04 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:55:02 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:55:57 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:56:52 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:57:49 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:58:47 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 22:59:41 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:00:37 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _____________________________ test_both_mergetree ______________________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_both_mergetree(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:104: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (139793411016256) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-03 23:01:20 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:01:33 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:01:33 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:01:33 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:01:34 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:02:31 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:03:26 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:04:21 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:05:18 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:06:16 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:07:11 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:08:06 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:09:03 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:10:01 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:10:55 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:11:50 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:12:47 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:13:45 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:14:40 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:15:35 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _______________________ test_not_work_on_different_disk ________________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_not_work_on_different_disk(start_cluster): cleanup([replica1, replica2]) # Replace and move should not work on replace > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:197: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (139793411016256) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-03 23:16:20 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:16:32 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:16:32 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:16:32 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:16:32 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:17:30 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:18:25 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:19:20 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:20:17 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:21:15 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:22:09 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:23:04 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:24:01 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:24:59 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:25:54 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:26:49 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:27:46 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:28:44 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:29:38 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:30:34 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _______________________ test_only_destination_replicated _______________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_only_destination_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:161: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (139793411016256) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-03 23:31:21 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:31:30 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:31:30 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:31:31 [ 680 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:31:31 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:32:29 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:33:23 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:34:18 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:35:15 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:36:13 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:37:08 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:38:03 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:39:00 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:39:58 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:40:52 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:41:48 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:42:45 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:43:42 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:44:37 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:45:32 [ 680 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) --------------------------- Captured stderr teardown --------------------------- Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml stop --timeout 20] Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml down --volumes] Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removed Stderr: Network roottestattachpartitionusingcopy-gw6_default Removing Stderr: Network roottestattachpartitionusingcopy-gw6_default Removed Cleanup called Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestattachpartitionusingcopy-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 ---------------------------- Captured log teardown ----------------------------- 2025-04-03 23:46:21 [ 680 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml stop --timeout 20] (cluster.py:122, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:122, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:122, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml down --volumes] (cluster.py:122, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removing (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removing (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removed (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removed (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removing (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removing (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removing (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removed (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removed (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removed (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Removing (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Removed (cluster.py:148, run_and_check) 2025-04-03 23:46:28 [ 680 ] DEBUG : Cleanup called (cluster.py:894, cleanup) 2025-04-03 23:46:28 [ 680 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-03 23:46:28 [ 680 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-03 23:46:29 [ 680 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-03 23:46:29 [ 680 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:122, run_and_check) 2025-04-03 23:46:29 [ 680 ] DEBUG : Unstopped containers: {} (cluster.py:908, cleanup) 2025-04-03 23:46:29 [ 680 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy-gw6 (cluster.py:922, cleanup) 2025-04-03 23:46:29 [ 680 ] DEBUG : Trying to prune unused networks... (cluster.py:928, cleanup) 2025-04-03 23:46:29 [ 680 ] DEBUG : Trying to prune unused images... (cluster.py:944, cleanup) 2025-04-03 23:46:29 [ 680 ] DEBUG : Command:[docker image prune -f] (cluster.py:122, run_and_check) 2025-04-03 23:46:29 [ 680 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:146, run_and_check) 2025-04-03 23:46:29 [ 680 ] DEBUG : Images pruned (cluster.py:947, cleanup) 2025-04-03 23:46:29 [ 680 ] DEBUG : Trying to prune unused volumes... (cluster.py:953, cleanup) 2025-04-03 23:46:29 [ 680 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:122, run_and_check) 2025-04-03 23:46:29 [ 680 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-03 23:46:29 [ 680 ] DEBUG : Volumes pruned: 1 (cluster.py:958, cleanup) ============================== slowest durations =============================== 900.00s call test_attach_partition_using_copy/test.py::test_only_destination_replicated 900.00s call test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 900.00s call test_attach_partition_using_copy/test.py::test_both_mergetree 882.22s call test_attach_partition_using_copy/test.py::test_all_replicated 174.89s setup test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 78.48s call test_ddl_worker_replicas/test.py::test_ddl_worker_replicas 51.63s call test_async_insert_memory/test.py::test_memory_usage 49.16s call test_dictionaries_dependency/test.py::test_no_lazy_load 33.44s call test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] 33.43s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] 32.48s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] 32.47s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] 32.44s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] 32.17s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] 31.41s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] 30.84s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] 30.43s call test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] 23.07s setup test_cluster_all_replicas/test.py::test_cluster 21.79s teardown test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation 21.48s teardown test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 21.32s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 21.05s call test_concurrent_queries_restriction_by_query_kind/test.py::test_select 18.84s setup test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation 18.67s call test_config_yaml_main/test.py::test_yaml_main_conf 18.66s setup test_compression_nested_columns/test.py::test_nested_compression_codec 18.05s call test_disk_access_storage/test.py::test_alter 17.78s setup test_attach_partition_using_copy/test.py::test_all_replicated 17.21s setup test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica 17.17s teardown test_backup_restore_on_cluster/test.py::test_tables_dependency 17.13s setup test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] 17.06s setup test_always_fetch_merged/test.py::test_replica_always_download 16.81s setup test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field 16.78s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] 16.30s call test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 15.53s setup test_backup_restore/test.py::test_attach_partition 15.12s call test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] 15.08s teardown test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica 14.85s setup test_accept_invalid_certificate/test.py::test_accept 14.58s call test_config_xml_main/test.py::test_xml_main_conf 14.51s setup test_default_role/test.py::test_alter_user 14.15s call test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] 13.88s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] 13.30s call test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] 13.27s setup test_disk_access_storage/test.py::test_alter 13.24s call test_config_decryption/test_wrong_settings.py::test_invalid_chars 13.16s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] 12.97s call test_config_decryption/test_wrong_settings.py::test_wrong_method 12.88s call test_config_decryption/test_wrong_settings.py::test_no_encryption_key 12.81s setup test_create_query_constraints/test.py::test_create_query_const_constraints 12.20s call test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] 12.15s setup test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] 11.78s setup test_ddl_worker_replicas/test.py::test_ddl_worker_replicas 10.92s call test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] 10.89s call test_config_decryption/test_wrong_settings.py::test_subnodes 10.57s call test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] 10.30s call test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 10.19s call test_always_fetch_merged/test.py::test_replica_always_download 9.86s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 8.73s setup test_custom_settings/test.py::test_custom_settings 8.27s call test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] 8.20s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 7.98s call test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] 7.84s teardown test_attach_partition_using_copy/test.py::test_only_destination_replicated 7.30s call test_backup_restore/test.py::test_attach_partition 6.95s setup test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 6.93s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 6.58s call test_backup_restore/test.py::test_replace_partition 6.55s teardown test_dictionaries_dependency/test.py::test_no_lazy_load 6.46s call test_backup_restore_on_cluster/test.py::test_replicated_database_async 6.23s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 6.18s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 5.80s call test_backup_restore_on_cluster/test.py::test_tables_dependency 5.77s call test_backup_restore/test.py::test_restore 5.53s teardown test_compression_nested_columns/test.py::test_nested_compression_codec 5.44s teardown test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field 5.32s call test_backup_restore_on_cluster/test.py::test_mutation 5.30s teardown test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] 5.25s teardown test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] 5.22s teardown test_ddl_worker_replicas/test.py::test_ddl_worker_replicas 5.19s teardown test_always_fetch_merged/test.py::test_replica_always_download 5.06s call test_backup_restore_on_cluster/test.py::test_replicated_database 5.01s call test_backup_restore_on_cluster/test.py::test_required_privileges 4.90s teardown test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] 4.84s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] 4.75s teardown test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] 4.61s teardown test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] 4.60s teardown test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] 4.60s teardown test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] 4.59s call test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 4.59s teardown test_backup_restore/test.py::test_restore 4.55s teardown test_concurrent_queries_restriction_by_query_kind/test.py::test_select 4.48s teardown test_create_query_constraints/test.py::test_create_query_minmax_constraints 4.48s teardown test_accept_invalid_certificate/test.py::test_strict_reject_with_config 4.34s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 4.30s teardown test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] 4.23s teardown test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] 4.23s teardown test_async_insert_memory/test.py::test_memory_usage 4.07s call test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field 4.05s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 4.02s call test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 3.76s teardown test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] 3.71s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 3.64s call test_backup_restore_on_cluster/test.py::test_replicated_table 3.64s call test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 3.60s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 3.60s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 3.44s call test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica 3.40s call test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation 3.33s call test_backup_restore_on_cluster/test.py::test_system_functions 3.33s call test_backup_restore_on_cluster/test.py::test_system_users 3.32s call test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 3.28s teardown test_default_role/test.py::test_wrong_set_default_role 3.27s setup test_async_insert_memory/test.py::test_memory_usage 3.25s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 3.09s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 3.06s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 3.02s teardown test_custom_settings/test.py::test_illformed_setting 3.02s call test_backup_restore_on_cluster/test.py::test_projection 3.01s call test_backup_restore_on_cluster/test.py::test_file_deduplication 3.00s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 2.91s call test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 2.80s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 2.78s call test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] 2.75s call test_default_role/test.py::test_set_default_roles 2.54s call test_create_query_constraints/test.py::test_create_query_minmax_constraints 2.50s call test_compression_nested_columns/test.py::test_nested_compression_codec 2.37s call test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] 2.32s call test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] 2.29s call test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 2.20s call test_create_query_constraints/test.py::test_create_query_const_constraints 2.19s call test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 2.06s call test_backup_restore_on_cluster/test.py::test_empty_replicated_table 1.92s call test_default_role/test.py::test_alter_user 1.86s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 1.70s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 1.58s teardown test_backup_restore_on_cluster/test.py::test_replicated_database 1.58s teardown test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 1.53s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 1.53s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_async 1.53s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 1.48s teardown test_backup_restore_on_cluster/test.py::test_mutation 1.48s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 1.48s teardown test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 1.43s teardown test_backup_restore_on_cluster/test.py::test_replicated_table 1.43s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 1.43s teardown test_backup_restore_on_cluster/test.py::test_empty_replicated_table 1.43s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 1.43s teardown test_backup_restore_on_cluster/test.py::test_file_deduplication 1.38s teardown test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 1.38s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 1.38s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 1.38s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 1.38s teardown test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 1.33s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 1.33s teardown test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 1.33s teardown test_backup_restore_on_cluster/test.py::test_projection 1.33s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 1.33s teardown test_backup_restore_on_cluster/test.py::test_system_users 1.33s teardown test_backup_restore_on_cluster/test.py::test_required_privileges 1.33s teardown test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 1.33s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 1.33s teardown test_backup_restore_on_cluster/test.py::test_system_functions 1.33s teardown test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 1.15s teardown test_disk_access_storage/test.py::test_alter 1.03s call test_default_role/test.py::test_wrong_set_default_role 0.99s call test_custom_settings/test.py::test_custom_settings 0.96s call test_cluster_all_replicas/test.py::test_cluster 0.75s call test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 0.65s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] 0.62s setup test_default_role/test.py::test_wrong_set_default_role 0.60s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] 0.60s call test_cluster_all_replicas/test.py::test_global_in 0.55s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] 0.55s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] 0.55s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] 0.32s call test_custom_settings/test.py::test_illformed_setting 0.22s call test_accept_invalid_certificate/test.py::test_connection_accept 0.22s call test_accept_invalid_certificate/test.py::test_accept 0.17s call test_accept_invalid_certificate/test.py::test_strict_reject_with_config 0.17s call test_accept_invalid_certificate/test.py::test_default 0.17s setup test_default_role/test.py::test_set_default_roles 0.17s call test_accept_invalid_certificate/test.py::test_strict_connection_reject 0.17s call test_accept_invalid_certificate/test.py::test_strict_reject 0.05s setup test_config_decryption/test_wrong_settings.py::test_invalid_chars 0.00s setup test_config_yaml_main/test.py::test_yaml_main_conf 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 0.00s setup test_config_xml_main/test.py::test_xml_main_conf 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] 0.00s teardown test_config_decryption/test_wrong_settings.py::test_no_encryption_key 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 0.00s setup test_backup_restore_on_cluster/test.py::test_system_users 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 0.00s setup test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 0.00s setup test_accept_invalid_certificate/test.py::test_strict_connection_reject 0.00s setup test_accept_invalid_certificate/test.py::test_strict_reject_with_config 0.00s setup test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 0.00s setup test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 0.00s setup test_custom_settings/test.py::test_illformed_setting 0.00s setup test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_mutation 0.00s setup test_backup_restore_on_cluster/test.py::test_empty_replicated_table 0.00s setup test_backup_restore_on_cluster/test.py::test_projection 0.00s setup test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 0.00s setup test_backup_restore_on_cluster/test.py::test_tables_dependency 0.00s teardown test_accept_invalid_certificate/test.py::test_default 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] 0.00s teardown test_config_xml_main/test.py::test_xml_main_conf 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_default_role/test.py::test_set_default_roles 0.00s setup test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 0.00s teardown test_custom_settings/test.py::test_custom_settings 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] 0.00s setup test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 0.00s setup test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 0.00s setup test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 0.00s teardown test_config_decryption/test_wrong_settings.py::test_wrong_method 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] 0.00s teardown test_config_yaml_main/test.py::test_yaml_main_conf 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] 0.00s setup test_accept_invalid_certificate/test.py::test_strict_reject 0.00s setup test_backup_restore_on_cluster/test.py::test_required_privileges 0.00s setup test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] 0.00s setup test_dictionaries_dependency/test.py::test_no_lazy_load 0.00s setup test_backup_restore_on_cluster/test.py::test_system_functions 0.00s setup test_config_decryption/test_wrong_settings.py::test_subnodes 0.00s setup test_backup_restore_on_cluster/test.py::test_file_deduplication 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] 0.00s setup test_config_decryption/test_wrong_settings.py::test_wrong_method 0.00s setup test_backup_restore/test.py::test_replace_partition 0.00s setup test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 0.00s setup test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] 0.00s teardown test_backup_restore/test.py::test_attach_partition 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] 0.00s setup test_accept_invalid_certificate/test.py::test_default 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] 0.00s teardown test_attach_partition_using_copy/test.py::test_all_replicated 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 0.00s setup test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] 0.00s teardown test_cluster_all_replicas/test.py::test_global_in 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 0.00s teardown test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_async 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table 0.00s teardown test_config_decryption/test_wrong_settings.py::test_subnodes 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] 0.00s teardown test_cluster_all_replicas/test.py::test_cluster 0.00s teardown test_default_role/test.py::test_alter_user 0.00s setup test_create_query_constraints/test.py::test_create_query_minmax_constraints 0.00s teardown test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] 0.00s teardown test_create_query_constraints/test.py::test_create_query_const_constraints 0.00s setup test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s setup test_backup_restore/test.py::test_restore 0.00s teardown test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_accept_invalid_certificate/test.py::test_strict_connection_reject 0.00s setup test_concurrent_queries_restriction_by_query_kind/test.py::test_select 0.00s setup test_accept_invalid_certificate/test.py::test_connection_accept 0.00s teardown test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s teardown test_accept_invalid_certificate/test.py::test_connection_accept 0.00s teardown test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] 0.00s setup test_cluster_all_replicas/test.py::test_global_in 0.00s setup test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s teardown test_accept_invalid_certificate/test.py::test_strict_reject 0.00s teardown test_config_decryption/test_wrong_settings.py::test_invalid_chars 0.00s teardown test_accept_invalid_certificate/test.py::test_accept 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] 0.00s setup test_config_decryption/test_wrong_settings.py::test_no_encryption_key 0.00s setup test_attach_partition_using_copy/test.py::test_only_destination_replicated 0.00s teardown test_backup_restore/test.py::test_replace_partition 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] =========================== short test summary info ============================ FAILED test_attach_partition_using_copy/test.py::test_all_replicated - Failed... FAILED test_attach_partition_using_copy/test.py::test_both_mergetree - Failed... FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated PASSED test_config_decryption/test_wrong_settings.py::test_invalid_chars PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] PASSED test_accept_invalid_certificate/test.py::test_accept PASSED test_accept_invalid_certificate/test.py::test_connection_accept PASSED test_accept_invalid_certificate/test.py::test_default PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] PASSED test_accept_invalid_certificate/test.py::test_strict_connection_reject PASSED test_accept_invalid_certificate/test.py::test_strict_reject PASSED test_accept_invalid_certificate/test.py::test_strict_reject_with_config PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] PASSED test_default_role/test.py::test_alter_user PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] PASSED test_default_role/test.py::test_set_default_roles PASSED test_default_role/test.py::test_wrong_set_default_role PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] PASSED test_backup_restore/test.py::test_attach_partition PASSED test_cluster_all_replicas/test.py::test_cluster PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] PASSED test_config_decryption/test_wrong_settings.py::test_no_encryption_key PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] PASSED test_backup_restore/test.py::test_replace_partition PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] PASSED test_backup_restore/test.py::test_restore PASSED test_config_decryption/test_wrong_settings.py::test_subnodes PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_insert PASSED test_custom_settings/test.py::test_custom_settings PASSED test_cluster_all_replicas/test.py::test_global_in PASSED test_custom_settings/test.py::test_illformed_setting PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica PASSED test_create_query_constraints/test.py::test_create_query_const_constraints PASSED test_create_query_constraints/test.py::test_create_query_minmax_constraints PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication PASSED test_config_decryption/test_wrong_settings.py::test_wrong_method PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_select PASSED test_backup_restore_on_cluster/test.py::test_mutation PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] PASSED test_always_fetch_merged/test.py::test_replica_always_download PASSED test_backup_restore_on_cluster/test.py::test_projection PASSED test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field PASSED test_backup_restore_on_cluster/test.py::test_replicated_database PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] PASSED test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts PASSED test_config_xml_main/test.py::test_xml_main_conf PASSED test_compression_nested_columns/test.py::test_nested_compression_codec PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] PASSED test_async_insert_memory/test.py::test_memory_usage PASSED test_backup_restore_on_cluster/test.py::test_replicated_table PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def PASSED test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] PASSED test_backup_restore_on_cluster/test.py::test_required_privileges PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] PASSED test_backup_restore_on_cluster/test.py::test_system_functions PASSED test_backup_restore_on_cluster/test.py::test_system_users PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] PASSED test_ddl_worker_replicas/test.py::test_ddl_worker_replicas PASSED test_dictionaries_dependency/test.py::test_no_lazy_load PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] PASSED test_config_yaml_main/test.py::test_yaml_main_conf PASSED test_disk_access_storage/test.py::test_alter ================== 4 failed, 96 passed in 3610.69s (1:00:10) =================== Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 528, in subprocess.check_call(cmd, shell=True) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_sbt52f --privileged --dns-search='.' --memory=30709035008 --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=6712d5cc610d -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e CLICKHOUSE_USE_OLD_ANALYZER=1 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_accept_invalid_certificate/test.py::test_accept test_accept_invalid_certificate/test.py::test_connection_accept test_accept_invalid_certificate/test.py::test_default test_accept_invalid_certificate/test.py::test_strict_connection_reject test_accept_invalid_certificate/test.py::test_strict_reject test_accept_invalid_certificate/test.py::test_strict_reject_with_config test_always_fetch_merged/test.py::test_replica_always_download test_async_insert_memory/test.py::test_memory_usage test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_compression_nested_columns/test.py::test_nested_compression_codec test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_config_decryption/test_wrong_settings.py::test_invalid_chars test_config_decryption/test_wrong_settings.py::test_no_encryption_key test_config_decryption/test_wrong_settings.py::test_subnodes test_config_decryption/test_wrong_settings.py::test_wrong_method test_config_xml_main/test.py::test_xml_main_conf test_config_yaml_main/test.py::test_yaml_main_conf test_create_query_constraints/test.py::test_create_query_const_constraints test_create_query_constraints/test.py::test_create_query_minmax_constraints test_custom_settings/test.py::test_custom_settings test_custom_settings/test.py::test_illformed_setting test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica test_ddl_worker_replicas/test.py::test_ddl_worker_replicas test_default_role/test.py::test_alter_user test_default_role/test.py::test_set_default_roles test_default_role/test.py::test_wrong_set_default_role 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1]' test_dictionaries_dependency/test.py::test_no_lazy_load 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache]' test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation test_disk_access_storage/test.py::test_alter -vvv" altinityinfra/integration-tests-runner:cd6390247eca ' returned non-zero exit status 1.